negative data
Positive-Unlabeled Learning using Random Forests via Recursive Greedy Risk Minimization
The need to learn from positive and unlabeled data, or PU learning, arises in many applications and has attracted increasing interest. While random forests are known to perform well on many tasks with positive and negative data, recent PU algorithms are generally based on deep neural networks, and the potential of tree-based PU learning is under-explored. In this paper, we propose new random forest algorithms for PU-learning. Key to our approach is a new interpretation of decision tree algorithms for positive and negative data as \emph{recursive greedy risk minimization algorithms}. We extend this perspective to the PU setting to develop new decision tree learning algorithms that directly minimizes PU-data based estimators for the expected risk. This allows us to develop an efficient PU random forest algorithm, PU extra trees. Our approach features three desirable properties: it is robust to the choice of the loss function in the sense that various loss functions lead to the same decision trees; it requires little hyperparameter tuning as compared to neural network based PU learning; it supports a feature importance that directly measures a feature's contribution to risk minimization. Our algorithms demonstrate strong performance on several datasets.
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- North America > Canada > Ontario > Toronto (0.04)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- North America > Canada > Ontario > Toronto (0.04)
We sincerely thank the three reviewers for their constructive comments and supports
We sincerely thank the three reviewers for their constructive comments and supports. GPUs are essential to doing effective deep learning. Cloud platform is used, no compression is required. The novelty of this paper is twofold. Secondly, we enhance the robustness of knowledge distillation to deal with data imbalance problem and noise.
- North America > United States > Virginia (0.04)
- North America > United States > Washington > King County > Seattle (0.04)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- (2 more...)
- Workflow (1.00)
- Research Report > Experimental Study (0.93)
- Education (0.46)
- Information Technology (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Europe > Spain > Andalusia > Granada Province > Granada (0.04)
- Asia > Middle East > Jordan (0.04)
- Information Technology (0.93)
- Health & Medicine > Therapeutic Area (0.46)
- Europe > Spain > Andalusia > Granada Province > Granada (0.04)
- Asia > Middle East > Jordan (0.04)
- Information Technology (0.93)
- Health & Medicine > Therapeutic Area (0.46)
- North America > United States > Virginia (0.04)
- North America > United States > Washington > King County > Seattle (0.04)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- (2 more...)
- Workflow (1.00)
- Research Report > Experimental Study (0.93)
- Education (0.67)
- Information Technology (0.46)
Export Reviews, Discussions, Author Feedback and Meta-Reviews
First provide a summary of the paper, and then address the following criteria: Quality, clarity, originality and significance. Summary: The paper presents two novel convolutional neural network architectures for modeling sentences in natural languages. These networks are trained specifically for the problem of matching a pair of sentences. The first architecture is a minor modification to the standard way of using a convolutional network over natural language sentences. After a convolution operation on the word embeddings, instead of doing a pooling operation across time (full sequence of words in a sentence) to select a single feature (or k features), the proposed model applies pooling to features associated with consecutive pairs of words.
Accessible, Realistic, and Fair Evaluation of Positive-Unlabeled Learning Algorithms
Wang, Wei, Wu, Dong-Dong, Li, Ming, Zhang, Jingxiong, Niu, Gang, Sugiyama, Masashi
Positive-unlabeled (PU) learning is a weakly supervised binary classification problem, in which the goal is to learn a binary classifier from only positive and unlabeled data, without access to negative data. In recent years, many PU learning algorithms have been developed to improve model performance. However, experimental settings are highly inconsistent, making it difficult to identify which algorithm performs better. In this paper, we propose the first PU learning benchmark to systematically compare PU learning algorithms. During our implementation, we identify subtle yet critical factors that affect the realistic and fair evaluation of PU learning algorithms. On the one hand, many PU learning algorithms rely on a validation set that includes negative data for model selection. This is unrealistic in traditional PU learning settings, where no negative data are available. To handle this problem, we systematically investigate model selection criteria for PU learning. On the other hand, the problem settings and solutions of PU learning have different families, i.e., the one-sample and two-sample settings. However, existing evaluation protocols are heavily biased towards the one-sample setting and neglect the significant difference between them. We identify the internal label shift problem of unlabeled training data for the one-sample setting and propose a simple yet effective calibration approach to ensure fair comparisons within and across families. We hope our framework will provide an accessible, realistic, and fair environment for evaluating PU learning algorithms in the future.
- North America > Canada > Ontario > Toronto (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- Asia > China (0.04)